Nonmonotone Line Search for Minimax Problems
نویسنده
چکیده
It was recently shown that, in the solution of smooth constrained optimization problems by sequential quadratic programming (SQP), the Maratos eeect can be prevented by means of a certain nonmonotone (more precisely, three-step or four-step monotone) line search. Using a well known transformation, this scheme can be readily extended to the case of minimax problems. It turns out however that, due to the structure of these problems, one can use a simpler scheme. Such a scheme is proposed and analyzed in this paper. Numerical experiments indicate a signiicant advantage of the proposed line search over the (monotone) Armijo search.
منابع مشابه
Avoiding the Maratos Effect by Means of a Nonmonotone Line Search Ii. Inequality Constrained Problems { Feasible Iterates*
When solving inequality constrained optimization problems via Sequential Quadratic Programming (SQP), it is potentially advantageous to generate iterates that all satisfy the constraints: all quadratic programs encountered are then feasible and there is no need for a surrogate merit function. (Feasibility of the successive iterates is in fact required in many contexts such as in real-time appli...
متن کاملA Nonmonotone Line Search Technique and Its Application to Unconstrained Optimization
A new nonmonotone line search algorithm is proposed and analyzed. In our scheme, we require that an average of the successive function values decreases, while the traditional nonmonotone approach of Grippo, Lampariello, and Lucidi [SIAM J. Numer. Anal., 23 (1986), pp. 707–716] requires that a maximum of recent function values decreases. We prove global convergence for nonconvex, smooth function...
متن کاملThe Global Convergence of Self-Scaling BFGS Algorithm with Nonmonotone Line Search for Unconstrained Nonconvex Optimization Problems
The self-scaling quasi-Newton method solves an unconstrained optimization problem by scaling the Hessian approximation matrix before it is updated at each iteration to avoid the possible large eigenvalues in the Hessian approximation matrices of the objective function. It has been proved in the literature that this method has the global and superlinear convergence when the objective function is...
متن کاملThe Global Convergence of Self-scale BFGS Algorithm with Nonmonotone Line Search for Unconstrained Nonconvex Optimization Problems
The self-scaling quasi-Newton method solves an unconstrained optimization problem by scaling the Hessian approximation matrix before it is updated at each iteration to avoid the possible large eigenvalues in the Hessian approximation matrices of the objective function. It has been proved in the literature that this method has the global and superlinear convergence when the objective function is...
متن کاملThe Riemannian Barzilai-borwein Method with Nonmonotone Line-search and the Matrix Geometric Mean Computation
The Barzilai-Borwein method, an effective gradient descent method with cleaver choice of the step-length, is adapted from nonlinear optimization to Riemannian manifold optimization. More generally, global convergence of a nonmonotone line-search strategy for Riemannian optimization algorithms is proved under some standard assumptions. By a set of numerical tests, the Riemannian Barzilai-Borwein...
متن کامل